Regularization with the Smooth-Lasso procedure
نویسنده
چکیده
We consider the linear regression problem. We propose the S-Lasso procedure to estimate the unknown regression parameters. This estimator enjoys sparsity of the representation while taking into account correlation between successive covariates (or predictors). The study covers the case when p ≫ n, i.e. the number of covariates is much larger than the number of observations. In the theoretical point of view, for fixed p, we establish asymptotic normality and consistency in variable selection results for our procedure. When p ≥ n, we provide variable selection consistency results and show that the S-Lasso achieved a Sparsity Inequality, i.e., a bound in term of the number of non-zero components of the oracle vector. It appears that the S-Lasso has nice variable selection properties compared to its challengers. Furthermore, we provide an estimator of the effective degree of freedom of the S-Lasso estimator. A simulation study shows that the S-Lasso performs better than the Lasso as far as variable selection is concerned especially when high correlations between successive covariates exist. This procedure also appears to be a good challenger to the Elastic-Net [36].
منابع مشابه
Smooth James-Stein model selection against erratic Stein unbiased risk estimate to select several regularization parameters
Smooth James-Stein thresholding-based estimators enjoy smoothness like ridge regression and perform variable selection like lasso. They have added flexibility thanks to more than one regularization parameters (like adaptive lasso), and the ability to select these parameters well thanks to a unbiased and smooth estimation of the risk. The motivation is a gravitational wave burst detection proble...
متن کاملNonlinear regression modeling and spike detection via Gaussian basis expansions
We consider the problem of constructing nonlinear regression models in the case that the structure of data has abrupt change points at unknown points. We propose two stage procedure where the spikes are detected by fused lasso signal approximator at the first stage, and the smooth curve are effectively estimated along with the technique of regularization method at the second. In order to select...
متن کاملVariable selection via the grouped weighted lasso for factor analysis models
The L1 regularization such as the lasso has been widely used in regression analysis since it tends to produce some coefficients that are exactly zero, which leads to variable selection. We consider the problem of variable selection for factor analysis models via the L1 regularization procedure. In order to select variables each of which is controlled by multiple parameters, we treat parameters ...
متن کاملModel-Consistent Sparse Estimation through the Bootstrap
We consider the least-square linear regression problem with regularization by the l 1-norm, a problem usually referred to as the Lasso. In this paper, we first present a detailed asymptotic analysis of model consistency of the Lasso in low-dimensional settings. For various decays of the regularization parameter, we compute asymptotic equivalents of the probability of correct model selection. Fo...
متن کاملActivity Identification and Local Linear Convergence of Forward-Backward-type Methods
In this paper, we consider a class of Forward–Backward (FB) splitting methods that includes several variants (e.g. inertial schemes, FISTA) for minimizing the sum of two proper convex and lower semi-continuous functions, one of which has a Lipschitz continuous gradient, and the other is partly smooth relatively to a smooth active manifoldM. We propose a unified framework, under which we show th...
متن کامل